Your browser doesn't support javascript.
Show: 20 | 50 | 100
Results 1 - 20 de 25
Filter
1.
13th IEEE Annual Computing and Communication Workshop and Conference, CCWC 2023 ; : 140-146, 2023.
Article in English | Scopus | ID: covidwho-2320850

ABSTRACT

Visualization is integral to investigating information hidden in data and providing users with intuitive feedback for decision-making. No matter the field a data set describes, inspecting the data visually will yield fruitful insights into the trends and statistics. Over the past calendar year, COVID-19 vaccines have become increasingly available for much of the population. However, the CDC (Centers for Disease Control and Prevention) fails to consider multiple sets of pandemic data in a side-by-side view and synchronize multiple key factors in one web page, limiting medical professionals and individuals to seeing, comparing, and interacting with complete data visualization. To analyze the coronavirus and vaccination data collected from multiple sources, effectively displaying them is critically important for interpreting the pandemic transmission pattern and vaccine efficiency. This paper presents new algorithms for innovative data visualizations that provide users with intuitive feedback and enable them to see a complete story of where the data is concerned. The information derived from our developed web-based data visualization will aid healthcare professionals and everyday citizens in moving forward as the pandemic progresses. © 2023 IEEE.

2.
Electronics ; 12(9):1977, 2023.
Article in English | ProQuest Central | ID: covidwho-2320345

ABSTRACT

Numerical information plays an important role in various fields such as scientific, financial, social, statistics, and news. Most prior studies adopt unsupervised methods by designing complex handcrafted pattern-matching rules to extract numerical information, which can be difficult to scale to the open domain. Other supervised methods require extra time, cost, and knowledge to design, understand, and annotate the training data. To address these limitations, we propose QuantityIE, a novel approach to extracting numerical information as structured representations by exploiting syntactic features of both constituency parsing (CP) and dependency parsing (DP). The extraction results may also serve as distant supervision for zero-shot model training. Our approach outperforms existing methods from two perspectives: (1) the rules are simple yet effective, and (2) the results are more self-contained. We further propose a numerical information retrieval approach based on QuantityIE to answer analytical queries. Experimental results on information extraction and retrieval demonstrate the effectiveness of QuantityIE in extracting numerical information with high fidelity.

3.
37th Annual Acm Symposium on Applied Computing ; : 813-820, 2022.
Article in English | Web of Science | ID: covidwho-2309179

ABSTRACT

In recent years, the trend of deploying digital systems in numerous industries has hiked. The health sector has observed an extensive adoption of digital systems and services that generate significant medical records. Electronic health records contain valuable information for prospective and retrospective analysis that is often not entirely exploited because of the complicated dense information storage. The crude purpose of condensing health records is to select the information that holds most characteristics of the original documents based on a reported disease. These summaries may boost diagnosis and save a doctor's time during a saturated workload situation like the COVID-19 pandemic. In this paper, we are applying a multi-head attention-based mechanism to perform extractive summarization of meaningful phrases on clinical notes. Our method finds major sentences for a summary by correlating tokens, segments, and positional embeddings of sentences in a clinical note. The model outputs attention scores that are statistically transformed to extract critical phrases for visualization on the heat-mapping tool and for human use.

4.
3rd IEEE International Conference on Power, Electronics and Computer Applications, ICPECA 2023 ; : 859-863, 2023.
Article in English | Scopus | ID: covidwho-2306600

ABSTRACT

In recent years, Covid-19 is one of the major health challenges facing the human population. Due to the highly infectious nature of Covid-19 and the difficulty of detecting symptoms in the early stages, it is definitely necessary to combine X-ray for the diagnosis of pneumonia. Using traditional neural networks such as VGG, ResNet, and DenseNet to diagnose pneumonia based on X-ray images faces a number of difficulties. These models have insufficient spatial information extraction capability and are prone to overfitting on the training set. The attention mechanism is a means to improve model performance by helping the model better extract channel and spatial features from the feature maps. To identify pneumonia more accurately, we combined the ResNet network and CBAM attention mechanism to design the ResNet101-cbam model with a series of data augmentation methods as well as training strategies. We used the same approach to add attention mechanisms to ResNet50, ResNet101 and ResNet152 and tested their performance. The results show that ResNet101-cbam is the best performing model overall. It achieved a recall of 0.8205, a precision of 0.822, and an accuracy of 0.8285 on the test set, while the original pretrained ResNet101 had a precision of 0.7280 and an accuracy of 0.7644. Its performance were better than the more complex model: ResNet152-cbam, a little bit, but the training speed is improved by more than 25%. More importantly, the model with the added attention mechanism effectively overcomes the effects of positive and negative sample imbalance. The ResNet101-cbam model can be used as a medical aid, which can improve diagnostic efficiency and help us better deal with large-scale pneumonia epidemics. © 2023 IEEE.

5.
Big Data and Cognitive Computing ; 7(1), 2023.
Article in English | Scopus | ID: covidwho-2259143

ABSTRACT

The spread of fake news related to COVID-19 is an infodemic that leads to a public health crisis. Therefore, detecting fake news is crucial for an effective management of the COVID-19 pandemic response. Studies have shown that machine learning models can detect COVID-19 fake news based on the content of news articles. However, the use of biomedical information, which is often featured in COVID-19 news, has not been explored in the development of these models. We present a novel approach for predicting COVID-19 fake news by leveraging biomedical information extraction (BioIE) in combination with machine learning models. We analyzed 1164 COVID-19 news articles and used advanced BioIE algorithms to extract 158 novel features. These features were then used to train 15 machine learning classifiers to predict COVID-19 fake news. Among the 15 classifiers, the random forest model achieved the best performance with an area under the ROC curve (AUC) of 0.882, which is 12.36% to 31.05% higher compared to models trained on traditional features. Furthermore, incorporating BioIE-based features improved the performance of a state-of-the-art multi-modality model (AUC 0.914 vs. 0.887). Our study suggests that incorporating biomedical information into fake news detection models improves their performance, and thus could be a valuable tool in the fight against the COVID-19 infodemic. © 2023 by the authors.

6.
20th International Conference on Language Engineering, ESOLEC 2022 ; : 147-151, 2022.
Article in English | Scopus | ID: covidwho-2236066

ABSTRACT

In this work, the stochastic dispersion of novel coronavirus disease 2019 (COVID-19) at the borders between France and Italy has been considered using a multi-input multi-output stochastic model. The physical effects of wind, temperature and altitude have been investigated as these factors and physical relationships are stochastic in nature. Stochastic terms have also been included to take into account the turbulence effect, and the r and om nature of the above physical parameters considered. Then, a method is proposed to identify the developed model's order and parameters. The actual data has been used in the identification and prediction process as a reference. These data have been divided into two parts: The first part is used to calculate the stochastic parameters of the model which are used to predict the COVID-19 level, while the second part is used as a check data. The predicted results are in good agreement with the check data. © 2022 IEEE.

7.
Front Public Health ; 10: 1070870, 2022.
Article in English | MEDLINE | ID: covidwho-2199554

ABSTRACT

Background: The high infection rate, severe symptoms, and evolving aspects of the COVID-19 pandemic provide challenges for a variety of medical systems around the world. Automatic information retrieval from unstructured text is greatly aided by Natural Language Processing (NLP), the primary approach taken in this field. This study addresses COVID-19 mortality data from the intensive care unit (ICU) in Kuwait during the first 18 months of the pandemic. A key goal is to extract and classify the primary and intermediate causes of death from electronic health records (EHRs) in a timely way. In addition, comorbid conditions or concurrent diseases were retrieved and analyzed in relation to a variety of causes of mortality. Method: An NLP system using the Python programming language is constructed to automate the process of extracting primary and secondary causes of death, as well as comorbidities. The system is capable of handling inaccurate and messy data, this includes inadequate formats, spelling mistakes and mispositioned information. A machine learning decision trees method is used to classify the causes of death. Results: For 54.8% of the 1691 ICU patients we studied, septic shock or sepsis-related multiorgan failure was the leading cause of mortality. About three-quarters of patients die from acute respiratory distress syndrome (ARDS), a common intermediate cause of death. An arrhythmia (AF) disorder was determined to be the strongest predictor of intermediate cause of death, whether caused by ARDS or other causes. Conclusion: We created an NLP system to automate the extraction of causes of death and comorbidities from EHRs. Our method processes messy and erroneous data and classifies the primary and intermediate causes of death of COVID-19 patients. We advocate arranging the EHR with well-defined sections and menu-driven options to reduce incorrect forms.


Subject(s)
COVID-19 , Respiratory Distress Syndrome , Humans , Natural Language Processing , Electronic Health Records , Pandemics , COVID-19/epidemiology
8.
30th International Conference on Software, Telecommunications and Computer Networks, SoftCOM 2022 ; 2022.
Article in English | Scopus | ID: covidwho-2146140

ABSTRACT

The primary means of informing the population in modern society is through news portals. This paper analyses the characteristics and effects that such way of communication creates. The influence was studied in particular on an example of current global phenomenon 'vaccination' (hr. cijepljenje). The research method follows the CRISP-DM process adapted to the digitalized form of textual data. The analysed corpus, in the form of natural spoken language, was scraped from Croatian news portals. The subsequent processing extracts information from unstructured textual sources and provides valuable insights, like how much a particular topic is represented in the article. Modeling is based on the application of multiple text mining algorithms, like Words Cloud, Topic Modelling, Concordance and Sentiment Analysis. The implemented model produces indicators for objective information interpretation. The findings suggest that the portals associated the notion of vaccination with the COVID-19 pandemic. Furthermore, this term was often used in a political context. The words used and predominantly negative character of texts dealing with vaccination has led to the transmission of negative emotions to readers. A significant aspect of the study is the fact that it was conducted on the corpus of texts written in Croatian - a relatively small and morphologically complex language. © 2022 University of Split, FESB.

9.
Ieee Access ; 10:116808-116818, 2022.
Article in English | Web of Science | ID: covidwho-2136068

ABSTRACT

Due to the rapid growth of the Internet and the subsequent rise of social media users, sharing information has become more flexible than ever before. This unrestricted freedom has also led to an increase in fake news. During the Covid-19 outbreak, fake news spread globally, negatively affecting authorities' decisions and the health of individuals. As a result, governments, media agencies, and academics have established fact-checking units and developed automatic detection systems. Research approaches to verify the veracity of news focused largely on writing styles, propagation patterns, and building knowledge bases that serve as a reference for fact checking. However, little work has been done to assess the credibility of the source of the claim to be checked. This paper proposes a general framework for detecting fake news that uses external evidence to verify the veracity of online news in a multilingual setting. Search results from Google are used as evidence, and the claim is cross-checked with the top five search results. Additionally, we associate a vector of credibility scores with each evidence source based on the domain name and website reputation metrics. All of these components are combined to derive better predictions of the veracity of claims. Further, we analyze the claim-evidence entailment relationship and select supporting and refuting evidence to cross-check with the claim. The approach without selection components yields better detection performance. In this work, we consider as a case study Covid-19 related news. Our framework achieves an F1-score of 0.85 and 0.97 in distinguishing fake from true news on XFact and Constraint datasets respectively. With the achieved results, the proposed framework present a promising automatic fact checker for both early and late detection.

10.
28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, KDD 2022 ; : 4832-4833, 2022.
Article in English | Scopus | ID: covidwho-2020400

ABSTRACT

Exploring the vast amount of rapidly growing scientific text data is highly beneficial for real-world scientific discovery. However, scientific text mining is particularly challenging due to the lack of specialized domain knowledge in natural language context, complex sentence structures in scientific writing, and multi-modal representations of scientific knowledge. This tutorial presents a comprehensive overview of recent research and development on scientific text mining, focusing on the biomedical and chemistry domains. First, we introduce the motivation and unique challenges of scientific text mining. Then we discuss a set of methods that perform effective scientific information extraction, such as named entity recognition, relation extraction, and event extraction. We also introduce real-world applications such as textual evidence retrieval, scientific topic contrasting for drug discovery, and molecule representation learning for reaction prediction. Finally, we conclude our tutorial by demonstrating, on real-world datasets (COVID-19 and organic chemistry literature), how the information can be extracted and retrieved, and how they can assist further scientific discovery. We also discuss the emerging research problems and future directions for scientific text mining. © 2022 Owner/Author.

11.
International Journal of Information Technology & Decision Making ; : 1-47, 2022.
Article in English | Web of Science | ID: covidwho-2020351

ABSTRACT

In the last two years, we have seen a huge number of debates and discussions on COVID-19 in social media. Many authors have analyzed these debates on Facebook and Twitter, while very few ones have considered Reddit. In this paper, we focus on this social network and propose three approaches to extract information from posts on COVID-19 published in it. The first performs a semi-automatic and dynamic classification of Reddit posts. The second automatically constructs virtual subreddits, each characterized by homogeneous themes. The third automatically identifies virtual communities of users with homogeneous themes. The three approaches represent an advance over the past literature. In fact, the latter lacks studies regarding classification algorithms capable of outlining the differences among the thousands of posts on COVID-19 in Reddit. Analogously, it lacks approaches able to build virtual subreddits with homogeneous topics or virtual communities of users with common interests.

12.
BMC Med Inform Decis Mak ; 22(Suppl 2): 147, 2022 06 02.
Article in English | MEDLINE | ID: covidwho-1875008

ABSTRACT

BACKGROUND: Knowledges graphs (KGs) serve as a convenient framework for structuring knowledge. A number of computational methods have been developed to generate KGs from biomedical literature and use them for downstream tasks such as link prediction and question answering. However, there is a lack of computational tools or web frameworks to support the exploration and visualization of the KG themselves, which would facilitate interactive knowledge discovery and formulation of novel biological hypotheses. METHOD: We developed a web framework for Knowledge Graph Exploration and Visualization (KGEV), to construct and visualize KGs in five stages: triple extraction, triple filtration, metadata preparation, knowledge integration, and graph database preparation. The application has convenient user interface tools, such as node and edge search and filtering, data source filtering, neighborhood retrieval, and shortest path calculation, that work by querying a backend graph database. Unlike other KGs, our framework allows fast retrieval of relevant texts supporting the relationships in the KG, thus allowing human reviewers to judge the reliability of the knowledge extracted. RESULTS: We demonstrated a case study of using the KGEV framework to perform research on COVID-19. The COVID-19 pandemic resulted in an explosion of relevant literature, making it challenging to make full use of the vast and heterogenous sources of information. We generated a COVID-19 KG with heterogenous information, including literature information from the CORD-19 dataset, as well as other existing knowledge from eight data sources. We showed the utility of KGEV in three intuitive case studies to explore and query knowledge on COVID-19. A demo of this web application can be accessed at http://covid19nlp.wglab.org . Finally, we also demonstrated a turn-key adaption of the KGEV framework to study clinical phenotypic presentation of human diseases by Human Phenotype Ontology (HPO), illustrating the versatility of the framework. CONCLUSION: In an era of literature explosion, the KGEV framework can be applied to many emerging diseases to support structured navigation of the vast amount of newly published biomedical literature and other existing biological knowledge in various databases. It can be also used as a general-purpose tool to explore and query gene-phenotype-disease-drug relationships interactively.


Subject(s)
COVID-19 , Humans , Pandemics , Pattern Recognition, Automated , Phenotype , Reproducibility of Results
13.
37th ACM/SIGAPP Symposium on Applied Computing, SAC 2022 ; : 813-820, 2022.
Article in English | Scopus | ID: covidwho-1874703

ABSTRACT

In recent years, the trend of deploying digital systems in numerous industries has hiked. The health sector has observed an extensive adoption of digital systems and services that generate significant medical records. Electronic health records contain valuable information for prospective and retrospective analysis that is often not entirely exploited because of the complicated dense information storage. The crude purpose of condensing health records is to select the information that holds most characteristics of the original documents based on a reported disease. These summaries may boost diagnosis and save a doctor's time during a saturated workload situation like the COVID-19 pandemic. In this paper, we are applying a multi-head attention-based mechanism to perform extractive summarization of meaningful phrases on clinical notes. Our method finds major sentences for a summary by correlating tokens, segments, and positional embeddings of sentences in a clinical note. The model outputs attention scores that are statistically transformed to extract critical phrases for visualization on the heat-mapping tool and for human use. © 2022 ACM.

14.
21st International Conference on Intelligent Systems Design and Applications, ISDA 2021 ; 418 LNNS:508-517, 2022.
Article in English | Scopus | ID: covidwho-1787718

ABSTRACT

The COVID-19 pandemic has led to an unprecedented challenge to public health. It resulted in global efforts to understand, record, and alleviate the disease. This research serves the purpose of generating a relevant summary related to Coronavirus. The research uses the COVID-19 Open Research Dataset (CORD-19) provided by Allen Institute for AI. The dataset contains 236,336 academic full-text articles as of July 19, 2021. This paper introduces a web-based system to handle user questions over the Coronavirus full-text scholarly articles. The system periodically runs backend services to process such large amount article with basic Natural Language Processing (NLP) techniques that include tokenization, N-Grams extraction, and part-of-speech (PoS) tagging. It automatically identifies the keywords from the question and uses cosine similarity to summarize the associated content and present to the user. This research will possibly benefit researchers, health workers as well as other individuals. Moreover, the same service can be used to train with the datasets of different domains (e.g., education) to generate a relevant summary for other user groups (e.g., students). © 2022, The Author(s), under exclusive license to Springer Nature Switzerland AG.

15.
12th International Conference on Computing Communication and Networking Technologies, ICCCNT 2021 ; 2021.
Article in English | Scopus | ID: covidwho-1752394

ABSTRACT

The Coronavirus pandemic has called for extensive research in the medical discipline. Since such disease outbreaks are about life and death of the patients, the doctors' and biomedical scientists' time is crucial. Research documents are usually comprehensive, often consuming the readers' time. A solution to it would be to extract information from the research text resembling the most relevant parts of the original text so that their valuable time is saved. The problem of text summarization is to create a shortened piece of text that represents the most relevant information from a relatively larger piece of text. This paper aims to ease the burden of the doctors so won't have to read the extensive research documents by constructing a summary of the most relevant parts of a medical research paper. A text summarization algorithm always works by quantifying the sentences by some means and analyzing scores. We use the TF-IDF quantification which is a popular way to quantify sentences. We select sentences with a high score and exclude those with a lower score, compared to a threshold. In a medical research paper, several sentences might have a low score, but they might be important if they contain biomedical entities. We use a dataset which has been constructed upon biomedical and COVID-19 terminology to construct a much better summary than some existing tools. As new methods keep coming up, this simple, yet robust approach gives us an accuracy of over 78% for most CORD-19 research papers. © 2021 IEEE.

16.
3rd International Conference on Data and Information Sciences, ICDIS 2021 ; 318:341-350, 2022.
Article in English | Scopus | ID: covidwho-1718601

ABSTRACT

Objective: The objective of this paper is to analyze approved areas of medical research related to COVID-19 from the United Arab Emirates (UAE) and World Health Organization (WHO) in order to identify key topics and themes for these two entities. The paper attempts to understand the key focus areas of the government and private agencies for further medical research in response to COVID-19. Research Design and Methods: In view of availability of large volumes of documents and advancements in computing systems, text mining has emerged as a significant tool to analyze large volumes of unstructured data. For this paper, we have applied latent semantic analysis (LSA) and singular value decomposition for text clustering. Findings: The findings of terms analysis results show various focus areas of medical research communities for UAE and WHO. Nutrition is a key theme of research in UAE whereas alternative medicines or infection study emerged as key focus areas for WHO. Further analysis of topic modeling indicates that topics like pneumonia and prevention approach has been a focus of approved research for WHO. Contribution/Value Added: The study contributes to text mining literature by providing a framework for analyzing research or policy documents at country or organization level. This can help to understand the key themes in COVID-19 response by various countries and organizations and identify the focus areas for them. © 2022, The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

17.
2021 IEEE Cloud Summit, Cloud Summit 2021 ; : 7-12, 2021.
Article in English | Scopus | ID: covidwho-1707014

ABSTRACT

The coronavirus COVID-19 pandemic has become the center of concern worldwide and hence the focus of media attention. Checking the coronavirus-related news and updates has become a daily routine of everyone. Hence, news processing and analytics become key solutions to harvest the real value of this massive amount of news. This conscious growth of published news about COVID-19 makes it hard for a variety of audiences to navigate through, analyze, and select the most important news (e.g., relevant information about the pandemic, its evolution, the vital precautions, and the necessary interventions). This can be realized using current and emerging technologies including Cloud computing, Artificial Intelligence (AI) and Deep Learning (DL). In this paper, we propose a framework to analyze the massive amount of public Covid-19 media reports over the Cloud. This framework encompasses four modules, including text preprocessing, deep learning, and machine learning-based news information extraction, and recommendation. We conducted experiments to evaluate three modules of our framework and the results we have obtained prove that combining derived information from the news reports provides the policymakers, health authorities, and the public, a complete picture of the way this virus is proliferating. Analyzing this data swiftly is a powerful tool to provide imperative answers to questions that are relevant to public health. © 2021 IEEE.

18.
4th International Conference on Communication, Information and Computing Technology, ICCICT 2021 ; 2021.
Article in English | Scopus | ID: covidwho-1700436

ABSTRACT

Use of computer in spite of human in knowledge discovery and engineering is a trend from last decade. In situation of COVID-19, it is need of hour to use such computer assisted assessment tools which will reduce the work of teachers in education domain. In Literature various techniques were developed to assess the answers of objective types questions by using machine learning but very less work carried out on assessment of answer of descriptive type questions. This paper provides a solution for the assessment of answer of descriptive type questions where teacher need not to use any paper or pen but the computer imitates like as teacher and evaluate answer of students. The primary objective is to represent results of subjective answers in the form of pictorial representations using cosine similarity technique. Cosine similarity is one of similarity measures which find similarity between two objects irrespective of their size. The results of web-based application are sufficient enough to withdraw the traditional teacher centric assessment approach and develop a computer-aid solution which is beneficial for educational institutions. © 2021 IEEE.

19.
Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 ; : 654-663, 2021.
Article in English | Scopus | ID: covidwho-1679075

ABSTRACT

In the Chinese medical insurance industry, the assessor's role is essential and requires significant efforts to converse with the claimant. This is a highly professional job that involves many parts, such as identifying personal information, collecting related evidence, and making a final insurance report. Due to the coronavirus (COVID-19) pandemic, the previous offline insurance assessment has to be conducted online. However, for the junior assessor often lacking practical experience, it is not easy to quickly handle such a complex online procedure, yet this is important as the insurance company needs to decide how much compensation the claimant should receive based on the assessor's feedback. In order to promote assessors' work efficiency and speed up the overall procedure, in this paper, we propose a dialogue-based information extraction system that integrates advanced NLP technologies for medical insurance assessment. With the assistance of our system, the average time cost of the procedure is reduced from 55 minutes to 35 minutes, and the total human resources cost is saved 30% compared with the previous offline procedure. Until now, the system has already served thousands of online claim cases. © 2021 Association for Computational Linguistics

20.
Dissertation Abstracts International: Section B: The Sciences and Engineering ; 83(3-B):No Pagination Specified, 2022.
Article in English | APA PsycInfo | ID: covidwho-1652166

ABSTRACT

As data grows exponentially, the need for information sharing to make meaningful decisions is on the increase. Multiple organizations leading different roles in dynamic situations often need to share mission-dependent data securely. Examples include contact tracing for a contagious disease such as COVID-19, maritime search and rescue operations, or creating a collaborative bid for a contract. In such examples, the ability to access data may need to change dynamically, depending on the situation of a mission (e.g., whether a person tested positive for a disease, a ship is in distress, or a bid-offer with given properties needs to be created). Also, there is a need to know how reliable data from each source is and improve the trust of multiple information by automatically detecting discrepancies among data from multiple sources. The current work does not address such a dynamic environment. In addition, most current work concerns data cleaning tools that use rules to detect data quality issues and most of these tools can only detect non-semantic issues.I have made two major contributions in this dissertation. The first is a framework to enable situation-aware access control in a federated Data-as-a-Service architecture by using Semantic Web technologies. This framework allows distributed query rewriting and semantic reasoning that automatically adds situation-based constraints to ensure that users can only see results they can access. We validate the framework by applying it to two dynamic use cases: maritime search and rescue operations and contact tracing for surveillance of a contagious disease. The experimental results of two use cases of SAR mission and Covid-19 demonstrate the effectiveness of the proposed framework. Organizations that need to share sensitive data sets securely during dynamic, limited duration scenarios can quickly adopt the framework.A second contribution is a novel approach for semantic discrepancy detection on data from multiple sources when some of the sources have structured data and other sources contain unstructured data. The proposed approach has three novel aspects: 1) a transfer learning method to extract information as entities from unstructured data automatically;2) algorithms to automatically match extracted entities with column data of other data sources;3) algorithms to automatically detect semantic disparity between matched entities and column data across data sources. The experimental results on two events data sets and a disaster relief data set demonstrate the effectiveness and efficiency of the proposed methods. Multiple organizations collaborating to share data can quickly use the proposed methods. (PsycInfo Database Record (c) 2022 APA, all rights reserved)

SELECTION OF CITATIONS
SEARCH DETAIL